Learning a context-free task with a recurrent neural network: An analysis of stability
نویسنده
چکیده
Natural languages exhibit context-free properties such as center-embedded clauses. Recent research has sought a model that performs on these features with human-like inconsistencies, rather than like traditional discrete automata. This search has recently focussed on recurrent neural networks. It has been shown theoretically that recurrent networks are computationally as powerful as Turing machines (Siegelmann, 1993). However, the class of problems that recurrent networks can learn, and the solutions that the networks nd are still not known. Rodriguez, Wiles and Elman (1996) looked at the representation employed by a recurrent neural network to process the context-free language a n b n. They found that highly coordinated dynamical structures were required for the network to perform the task. This paper examines how a recurrent network learns these dynamics given the structural dependencies required. To observe the evolution of the networks, their weights and performance were monitored during training. We found that as a network approached the solution, the weights became unstable and any solution that was found was soon lost. An analysis of the changes in dynamics as the network loses a solution indicates that small changes in weights result in signiicant changes to the dynamics of the network. We conclude that learning context-free languages may be inherently unstable, but that further work is necessary to explore the possible reasons for this instability.
منابع مشابه
A Recurrent Neural Network Model for solving CCR Model in Data Envelopment Analysis
In this paper, we present a recurrent neural network model for solving CCR Model in Data Envelopment Analysis (DEA). The proposed neural network model is derived from an unconstrained minimization problem. In the theoretical aspect, it is shown that the proposed neural network is stable in the sense of Lyapunov and globally convergent to the optimal solution of CCR model. The proposed model has...
متن کاملA Recurrent Neural Network to Identify Efficient Decision Making Units in Data Envelopment Analysis
In this paper we present a recurrent neural network model to recognize efficient Decision Making Units(DMUs) in Data Envelopment Analysis(DEA). The proposed neural network model is derived from an unconstrained minimization problem. In theoretical aspect, it is shown that the proposed neural network is stable in the sense of lyapunov and globally convergent. The proposed model has a single-laye...
متن کاملMulti-Step-Ahead Prediction of Stock Price Using a New Architecture of Neural Networks
Modelling and forecasting Stock market is a challenging task for economists and engineers since it has a dynamic structure and nonlinear characteristic. This nonlinearity affects the efficiency of the price characteristics. Using an Artificial Neural Network (ANN) is a proper way to model this nonlinearity and it has been used successfully in one-step-ahead and multi-step-ahead prediction of di...
متن کاملUsing Prior Knowledge in an NNPDA to Learn Context-Free Languages
Although considerable interest has been shown in language inference and automata induction using recurrent neural networks, success of these models has mostly been limited to regular languages. We have previously demonstrated that Neural Network Pushdown Automaton (NNPDA) model is capable of learning deterministic context-free languages (e.g., a n b n and parenthesis languages) from examples. H...
متن کاملAn efficient one-layer recurrent neural network for solving a class of nonsmooth optimization problems
Constrained optimization problems have a wide range of applications in science, economics, and engineering. In this paper, a neural network model is proposed to solve a class of nonsmooth constrained optimization problems with a nonsmooth convex objective function subject to nonlinear inequality and affine equality constraints. It is a one-layer non-penalty recurrent neural network based on the...
متن کامل